Inferring dependencies in Embodiment-based modular reinforcement learning

نویسندگان

  • David Jacob
  • Daniel Polani
  • Chrystopher L. Nehaniv
چکیده

The state-spaces needed to describe realistic physical embodied agents are extremely large, which presents a serious challenge to classical reinforcement learning schemes. In previous work (Jacob et al., 2005a, Jacob et al., 2005b) we introduced our EMBER (for EMbodiment-Based modulaR) reinforcement learning system, which describes a novel method for decomposing agents into modules based on the agent’s embodiment. This modular decomposition factorises the statespace and dramatically improves performance in unknown and dynamic environments. However, while there are great advantages to be gained from a factorised state-space, the question of dependencies cannot be ignored. We present a development of the work reported in (Jacob et al., 2004) which shows, in a simple example, how dependencies may be identified using a heuristic approach. Results show that the system is able quickly to discover and act upon dependencies, even where they are neither simple nor deterministic.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Modular SRV Reinforcement Learning Architectures for Non-linear Control

This paper demonstrates the advantages of using a hybrid reinforcement–modular neural network architecture for non-linear control. Specifically, the method of ACTION-CRITIC reinforcement learning, modular neural networks, competitive learning and stochastic updating are combined. This provides an architecture able to both support temporal difference learning and probabilistic partitioning of th...

متن کامل

Reinforcement learning and causal models

This chapter reviews the diverse roles that causal knowledge plays in reinforcement learning. The first half of the chapter contrasts a “model-free” system that learns to repeat actions that lead to reward with a “model-based” system that learns a probabilistic causal model of the environment which it then uses to plan action sequences. Evidence suggests that these two systems coexist in the br...

متن کامل

Modular Neural Networks and Reinforcement Learning

We investigate the effect of modular architecture in an artificial neural network for a reinforcement learning problem. Using the supervised backpropagation algorithm to solve a two-task problem, the network performance can be increased by using networks with modular structures. However, using a modular architecture to solve a two-task reinforcement learning problem will not increase the perfor...

متن کامل

Dynamic Obstacle Avoidance by Distributed Algorithm based on Reinforcement Learning (RESEARCH NOTE)

In this paper we focus on the application of reinforcement learning to obstacle avoidance in dynamic Environments in wireless sensor networks. A distributed algorithm based on reinforcement learning is developed for sensor networks to guide mobile robot through the dynamic obstacles. The sensor network models the danger of the area under coverage as obstacles, and has the property of adoption o...

متن کامل

Inter-module credit assignment in modular reinforcement learning

Critical issues in modular or hierarchical reinforcement learning (RL) are (i) how to decompose a task into sub-tasks, (ii) how to achieve independence of learning of sub-tasks, and (iii) how to assure optimality of the composite policy for the entire task. The second and last requirements are often under trade-off. We propose a method for propagating the reward for the entire task achievement ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2005